产量估计是葡萄园管理中的强大工具,因为它允许种植者微调实践以优化产量和质量。但是,目前使用手动抽样进行估计,这是耗时和不精确的。这项研究表明,近端成像的应用与深度学习相结合,以进行葡萄园中的产量估计。使用车辆安装的传感套件进行连续数据收集,并使用商业收益率监控器在收获时结合了地面真实收益数据的收集,可以生成一个23,581个收益点和107,933张图像的大数据集。此外,这项研究是在机械管理的商业葡萄园中进行的,代表了一个充满挑战的图像分析环境,但在加利福尼亚中央山谷中的一组常见条件。测试了三个模型架构:对象检测,CNN回归和变压器模型。对象检测模型在手工标记的图像上进行了训练以定位葡萄束,并将束数量或像素区域求和以与葡萄产量相关。相反,回归模型端到端训练,以预测图像数据中的葡萄产量,而无需手动标记。结果表明,在代表性的保留数据集上,具有相当的绝对百分比误差为18%和18.5%的变压器和具有像素区域处理的对象检测模型。使用显着映射来证明CNN模型的注意力位于葡萄束的预测位置附近以及葡萄树冠的顶部。总体而言,该研究表明,近端成像和深度学习对于大规模预测葡萄群的适用性。此外,端到端建模方法能够与对象检测方法相当地执行,同时消除了手工标记的需求。
translated by 谷歌翻译
X射线微型计算机断层扫描(X射线Microct)已使以微米尺度上的植物和土壤中发生的特性和过程表征。尽管这种高级技术广泛使用,但硬件和软件的主要限制都限制了图像处理和数据分析的速度和准确性。机器学习的最新进展,特别是将卷积神经网络应用于图像分析的应用,已实现了图像数据的快速而准确的分割。然而,在将卷积神经网络应用于环境和农业相关图像的分析中仍然存在挑战。具体而言,计算机科学家和工程师,构建这些AI/ML工具的工程师与农业研究中潜在的最终用户之间存在脱节,他们可能不确定如何在其工作中应用这些工具。此外,与传统的计算系统相比,培训和应用深度学习模型所需的计算资源是独特的,对计算机游戏系统或图形设计工作更为常见。为了应对这些挑战,我们开发了一个模块化工作流程,用于使用Googles Colaboragoration Web应用程序中的低成本资源,将卷积神经网络应用于X射线Microct图像。在这里,我们介绍了工作流的结果,说明了如何使用核桃叶,杏仁花芽和土壤骨料的示例扫描来优化参数以获得最佳结果。我们预计该框架将加速植物和土壤科学中新兴的深度学习技术的采用和使用。
translated by 谷歌翻译
近年来,深度学习模型已成为农业计算机愿景的标准。这样的模型通常使用最初适合更通用的非农业数据集的模型权重对农业任务进行微调。缺乏农业特定的微调可能会增加训练时间和资源的使用,并降低模型性能,从而导致数据效率的总体下降。为了克服这一限制,我们为三个不同的任务收集了广泛的现有公共数据集,标准化它们,并构建标准培训和评估管道,为我们提供了一组基准测试和预处理的模型。然后,我们使用在深度学习任务中常用的方法进行了许多实验,但在其特定领域的农业应用中未探索。我们的实验指导我们开发多种方法,以提高培训农业深度学习模型,而没有对现有管道进行大规模修改。我们的结果表明,即使是使用农业预审预告额的模型权重,或将特定的空间增强量用于数据处理管道,也可以显着提高模型性能并导致较短的收敛时间,从而节省训练资源。此外,我们发现,即使是在低质量注释中训练的模型也可以产生与高质量等效物的可比性水平,这表明注释差的数据集仍然可以用于培训,扩大当前可用数据集的池。我们的方法在整个农业深度学习中广泛适用,并具有重大数据效率提高的高潜力。
translated by 谷歌翻译
超越地球轨道的人类空间勘探将涉及大量距离和持续时间的任务。为了有效减轻无数空间健康危害,数据和空间健康系统的范式转移是实现地球独立性的,而不是Earth-Reliance所必需的。有希望在生物学和健康的人工智能和机器学习领域的发展可以解决这些需求。我们提出了一个适当的自主和智能精密空间健康系统,可以监控,汇总和评估生物医学状态;分析和预测个性化不良健康结果;适应并响应新累积的数据;并提供对其船员医务人员的个人深度空间机组人员和迭代决策支持的预防性,可操作和及时的见解。在这里,我们介绍了美国国家航空航天局组织的研讨会的建议摘要,以便在太空生物学和健康中未来的人工智能应用。在未来十年,生物监测技术,生物标志科学,航天器硬件,智能软件和简化的数据管理必须成熟,并编织成精确的空间健康系统,以使人类在深空中茁壮成长。
translated by 谷歌翻译
空间生物学研究旨在了解太空飞行对生物的根本影响,制定支持深度空间探索的基础知识,最终生物工程航天器和栖息地稳定植物,农作物,微生物,动物和人类的生态系统,为持续的多行星寿命稳定。要提高这些目标,该领域利用了来自星空和地下模拟研究的实验,平台,数据和模型生物。由于研究扩展到低地球轨道之外,实验和平台必须是最大自主,光,敏捷和智能化,以加快知识发现。在这里,我们介绍了由美国国家航空航天局的人工智能,机器学习和建模应用程序组织的研讨会的建议摘要,这些应用程序为这些空间生物学挑战提供了关键解决方案。在未来十年中,将人工智能融入太空生物学领域将深化天空效应的生物学理解,促进预测性建模和分析,支持最大自主和可重复的实验,并有效地管理星载数据和元数据,所有目标使生活能够在深空中茁壮成长。
translated by 谷歌翻译
特征测量对于植物育种和农业生产管道至关重要。通常,使用费力的手动测量测量一套植物特征,然后用于培训和/或验证更高的吞吐量特征估计技术。这里,我们介绍了一种相对简单的卷积神经网络(CNN)模型,该模型接受多个传感器输入并预测多个连续特征输出 - 即多输入,多输出CNN(MIMO-CNN)。此外,我们将可变形的卷积层引入该网络架构(MIMO-DCNN),以使模型能够自适应地调整其接收领域,模拟数据中的复杂变量几何变换,以及微调连续的特征输出。我们检查MIMO-CNN和MIMO-DCNN模型如何在多输入(即RGB和深度图像)上执行,来自2021年自主温室挑战的多特征输出莴苣数据集。进行了消融研究以检查使用单一与多个输入的效果,以及单个与多个输出。 MIMO-DCNN模型导致归一化平均平方误差(NMSE)为0.068 - 顶部2021排行榜得分为0.081的实质性改进。提供了开源代码。
translated by 谷歌翻译
在数值天气和气候模型中的云结构的处理通常很大程度上是大大简化的,以使它们计算得起价格实惠。在这里,我们建议使用计算廉价的神经网络来纠正欧洲的中等天气预报1D辐射方案ECRAD,用于3D云效应。 3D云效应被学习为ECRAD快速1D Tripleclouds疏忽它们的差异及其3D Spartacus(通过云侧辐射传输的快速算法),其中包括它们的求解器,但大约是计算昂贵的五倍。在3D信号的20到30%之间的典型误差,神经网络的准确性提高了运行时增加约1%。因此,而不是模仿整个斯巴达斯,我们将Tripleclouds保持不变的气氛的无云部分和在其他地方的3D矫正它。如果我们假设两者的相似的信噪比,则对相对小的3D校正而不是整个信号的焦点允许显着提高预测。
translated by 谷歌翻译
We demonstrate a proof-of-concept of a large language model conducting corporate lobbying related activities. We use an autoregressive large language model (OpenAI's text-davinci-003) to determine if proposed U.S. Congressional bills are relevant to specific public companies and provide explanations and confidence levels. For the bills the model deems as relevant, the model drafts a letter to the sponsor of the bill in an attempt to persuade the congressperson to make changes to the proposed legislation. We use hundreds of ground-truth labels of the relevance of a bill to a company to benchmark the performance of the model, which outperforms the baseline of predicting the most common outcome of irrelevance. However, we test the ability to determine the relevance of a bill with the previous OpenAI GPT-3 model (text-davinci-002), which was state-of-the-art on many language tasks until text-davinci-003 was released on November 28, 2022. The performance of text-davinci-002 is worse than simply always predicting that a bill is irrelevant to a company. These results suggest that, as large language models continue to improve core natural language understanding capabilities, performance on corporate lobbying related tasks will continue to improve. We then discuss why this could be problematic for societal-AI alignment.
translated by 谷歌翻译
In the past years, deep learning has seen an increase of usage in the domain of histopathological applications. However, while these approaches have shown great potential, in high-risk environments deep learning models need to be able to judge their own uncertainty and be able to reject inputs when there is a significant chance of misclassification. In this work, we conduct a rigorous evaluation of the most commonly used uncertainty and robustness methods for the classification of Whole-Slide-Images under domain shift using the H\&E stained Camelyon17 breast cancer dataset. Although it is known that histopathological data can be subject to strong domain shift and label noise, to our knowledge this is the first work that compares the most common methods for uncertainty estimation under these aspects. In our experiments, we compare Stochastic Variational Inference, Monte-Carlo Dropout, Deep Ensembles, Test-Time Data Augmentation as well as combinations thereof. We observe that ensembles of methods generally lead to higher accuracies and better calibration and that Test-Time Data Augmentation can be a promising alternative when choosing an appropriate set of augmentations. Across methods, a rejection of the most uncertain tiles leads to a significant increase in classification accuracy on both in-distribution as well as out-of-distribution data. Furthermore, we conduct experiments comparing these methods under varying conditions of label noise. We observe that the border regions of the Camelyon17 dataset are subject to label noise and evaluate the robustness of the included methods against different noise levels. Lastly, we publish our code framework to facilitate further research on uncertainty estimation on histopathological data.
translated by 谷歌翻译
In large-scale machine learning, recent works have studied the effects of compressing gradients in stochastic optimization in order to alleviate the communication bottleneck. These works have collectively revealed that stochastic gradient descent (SGD) is robust to structured perturbations such as quantization, sparsification, and delays. Perhaps surprisingly, despite the surge of interest in large-scale, multi-agent reinforcement learning, almost nothing is known about the analogous question: Are common reinforcement learning (RL) algorithms also robust to similar perturbations? In this paper, we investigate this question by studying a variant of the classical temporal difference (TD) learning algorithm with a perturbed update direction, where a general compression operator is used to model the perturbation. Our main technical contribution is to show that compressed TD algorithms, coupled with an error-feedback mechanism used widely in optimization, exhibit the same non-asymptotic theoretical guarantees as their SGD counterparts. We then extend our results significantly to nonlinear stochastic approximation algorithms and multi-agent settings. In particular, we prove that for multi-agent TD learning, one can achieve linear convergence speedups in the number of agents while communicating just $\tilde{O}(1)$ bits per agent at each time step. Our work is the first to provide finite-time results in RL that account for general compression operators and error-feedback in tandem with linear function approximation and Markovian sampling. Our analysis hinges on studying the drift of a novel Lyapunov function that captures the dynamics of a memory variable introduced by error feedback.
translated by 谷歌翻译